| Prediction method | Details | File size limit |
|-------------------|---------|-----------------|
| Leaderboard predictions | To make predictions on a non-deployed model using the UI, expand the model on the Leaderboard and select [**Predict > Make Predictions**](predict). Upload predictions from a local file, URL, data source, or the AI Catalog. You can also upload predictions using the modeling predictions API, also called the "[V2 predictions API](https://datarobot-public-api-client.readthedocs-hosted.com/en/v2.26.1/entities/predict_job.html){ target=_blank }." Use this API to test predictions using your modeling workers on small datasets. Predictions can be limited to 100 requests per user, per hour, depending on your DataRobot package. | 1GB |
| Batch predictions (UI) | To make batch predictions using the UI, deploy a model and navigate to the deployment's  [**Make Predictions**](batch-pred) tab (requires MLOps). | 5GB |
| Batch predictions (API)  | The [Batch Prediction API](batch-prediction-api/index) is optimized for high-throughput and contains production grade connectivity options that allow you to not only push data through the API, but also connect to the AI catalog, cloud storage, databases, or data warehouses (requires MLOps). | Unlimited |
| Prediction API (real-time) | To make real-time predictions on a deployed model, use the [Prediction API](dr-predapi). | 50 MB |
